专利摘要:
A method for processing a video signal according to the present invention comprises: determining an intra prediction mode of a current block; inducing a reference sample for an intra prediction of the current block; and performing the intra prediction of the current block on the basis of the intra prediction mode and the reference sample.
公开号:ES2677193A2
申请号:ES201890011
申请日:2016-08-29
公开日:2018-07-30
发明作者:Bae Keun Lee;Joo Young Kim
申请人:KT Corp;
IPC主号:
专利说明:

5
10
fifteen
twenty
25
30
35
Procedure and device for processing video signals Technical field
The present invention relates to a method and a device for processing a video signal.
Background of the technique
Recently, the demand for high resolution and high quality images, such as high definition (HD) images and ultra high definition (UHD) images has increased in various fields of application. However, higher resolution and quality image data have increasing amounts of data compared to conventional image data. Therefore, when image data is transmitted using a medium such as conventional wired and wireless broadband networks, or when image data is stored using a conventional storage medium, transmission and storage costs increase. In order to solve these problems that occur with an increase in the resolution and quality of the image data, high efficiency image encoding / decoding techniques can be used.
Image compression technology includes various techniques, including: an inter prediction technique to predict a pixel value included in a current image from a previous or subsequent image of the current image; an intra prediction technique for predicting a pixel value included in a current image using the pixel information in the current image; an entropy codification technique for assigning a short code to a value with a high frequency of appearance and assigning a long code to a value with a low frequency of appearance; etc. Image data can be effectively compressed using said image compression technology, and can be transmitted or stored.
Meanwhile, with the demands for high-resolution images, the demands for stereographic image content, which is a new image service, have also increased. A video compression technique is being discussed to provide
5
10
fifteen
twenty
25
30
35
Divulgation
Technical problem
An object of the present invention is intended to provide a method and a device for encoding / decoding a video signal, partitioning the procedure and the device hierarchically a coding block.
An objective of the present invention is intended to provide a method and a device for encoding / decoding a video signal, the procedure and the device performing the intra prediction of an encoding / decoding target block.
An objective of the present invention is intended to provide a method and a device for encoding / decoding a video signal, correcting the procedure and the device a prediction sample of an objective block of encoding / decoding.
Technical solution
In accordance with the present invention, a method and a device for decoding a video signal are provided, including the procedure: determining an intra prediction mode of a current block; obtain a reference sample for the intra prediction of the current block; and perform the intra prediction of the current block based on the intra prediction mode and the reference sample.
In the method and the device for decoding a video signal in accordance with the present invention, the current block may be an encoding block in a non-square partitioned form based on at least one of between a square tree and a binary tree.
In the method and the device for decoding a video signal in accordance with the present invention, the determination of the intra prediction mode may include: generating a list of candidates having multiple candidates; and determine the intra prediction mode of the current block based on the list of candidates and an index information, the
3
5
10
fifteen
twenty
25
30
35
index information specifying a candidate index included in the list of candidates.
In the procedure and the device for decoding a video signal in accordance with the present invention, a maximum number of candidates that may be included in the list of candidates may be more than three.
In the method and the device for decoding a video signal in accordance with the present invention, the determined intra prediction mode may be one of the intra extended prediction modes, and the extended intra prediction modes may include a flat mode, a DC mode and more than 33 directional prediction modes.
In the method and the device for decoding a video signal in accordance with the present invention, the realization of the intra prediction may include: obtaining a prediction sample of the current block based on the intra prediction mode and the reference sample; and correct the prediction sample obtained using the differential information of the neighboring samples of the current block.
In the method and the device for decoding a video signal in accordance with the present invention, performing the intra prediction may include: obtaining a prediction sample of the current block based on the intra prediction mode and the reference sample; and correct the prediction sample obtained based on at least one of the predetermined weight and displacement.
In accordance with the present invention, there is provided a method and a device for encoding a video signal, including the procedure: determining an intra prediction mode of a current block; obtain a reference sample for the intra prediction of the current block; and perform the intra prediction of the current block based on the intra prediction mode and the reference sample.
In the method and the device for encoding a video signal according to the present invention, the current block may be a coding block in a non-square partitioned form based on at least one of between a square tree and a binary tree.
In the method and the device for encoding a video signal in accordance with the present invention, the determination of the intra prediction mode may include: generating a
4
5
10
fifteen
twenty
25
30
35
list of candidates that have multiple candidates; and determine the intra prediction mode of the current block based on the list of candidates and an index.
In the procedure and the device for encoding a video signal in accordance with the present invention, a maximum number of candidates that may be included in the list of candidates may be more than three.
In the method and the device for encoding a video signal in accordance with the present invention, the determined intra prediction mode may be one of the intra extended prediction modes, and the extended intra prediction modes include a flat mode, a mode DC and more than 33 directional prediction modes.
In the method and the device for encoding a video signal in accordance with the present invention, the realization of the intra prediction may include: obtaining a prediction sample of the current block based on the intra prediction mode and the reference sample; and correct the prediction sample obtained using the differential information of the neighboring samples of the current block.
In the method and the device for encoding a video signal in accordance with the present invention, performing the intra prediction may include: obtaining a prediction sample of the current block based on the intra prediction mode and the reference sample; and correct the prediction sample obtained based on at least one of the predetermined weight and displacement.
Advantageous effects
In accordance with the present invention, it is possible to improve the efficiency of codification through a hierarchical / adaptive partitioning of a coding block.
In accordance with the present invention, it is possible to effectively determine an intra prediction mode of an objective coding / decoding block, and improve the accuracy of the intra prediction.
Description of the drawings
Figure 1 is a block diagram illustrating a device for encoding a video of
5
5
10
fifteen
twenty
25
30
35
Fig. 2 is a block diagram illustrating a device for decoding a video in accordance with an embodiment of the present invention.
Figure 3 is a view illustrating an example of hierarchical partitioning of a codification block based on a tree structure according to an embodiment of the present invention.
Figure 4 is a view illustrating types of predefined intra prediction modes for a device for encoding / decoding a video in accordance with an embodiment of the present invention.
Fig. 5 is a flow chart that briefly illustrates an intra prediction procedure according to an embodiment of the present invention.
Figure 6 is a view illustrating a procedure for correcting a prediction sample of a current block based on the differential information of the neighboring samples according to an embodiment of the present invention.
Figures 7 and 8 are views illustrating a method for correcting a prediction sample based on a predetermined correction filter according to an embodiment of the present invention.
Fig. 9 is a view illustrating a procedure for correcting a prediction sample using weight and displacement according to an embodiment of the present invention.
Figures 10 to 15 are views illustrating a method of composing a template for determining the weight w in accordance with an embodiment of the present invention.
Better mode
In accordance with the present invention, a method and a device for decoding a video signal are provided, including the procedure: determining an intra prediction mode of a current block; get a reference sample for the prediction
6
5
10
fifteen
twenty
25
30
35
intra of the current block; and perform the intra prediction of the current block based on the intra prediction mode and the reference sample.
In the method and the device for decoding a video signal in accordance with the present invention, the current block may be an encoding block in a non-square partitioned form based on at least one of between a square tree and a binary tree.
In the method and the device for decoding a video signal in accordance with the present invention, the determination of the intra prediction mode may include: generating a list of candidates having multiple candidates; and determine the intra prediction mode of the current block based on the list of candidates and an index.
In the procedure and the device for decoding a video signal in accordance with the present invention, a maximum number of candidates that may be included in the list of candidates may be more than three.
In the method and the device for decoding a video signal in accordance with the present invention, the intra-prediction mode determined may be one of the intra-extended prediction modes, and the intra-extended prediction modes may include a flat mode, a DC mode and more than 33 directional prediction modes.
In the method and the device for decoding a video signal in accordance with the present invention, the realization of the intra prediction may include: obtaining a prediction sample of the current block based on the intra prediction mode and the reference sample; and correct the prediction sample obtained using the differential information of the neighboring samples of the current block.
In the method and the device for decoding a video signal in accordance with the present invention, performing the intra prediction may include: obtaining a prediction sample of the current block based on the intra prediction mode and the reference sample; and correct the prediction sample obtained based on at least one of the predetermined weight and displacement.
In accordance with the present invention, there is provided a method and a device for encoding a video signal, including the procedure: determining a mode of
7
5
10
fifteen
twenty
25
30
35
intra prediction of a current block; obtain a reference sample for the intra prediction of the current block; and perform the intra prediction of the current block based on the intra prediction mode and the reference sample.
In the method and the device for encoding a video signal in accordance with the present invention, the current block may be a codification block in a non-square partitioned form based on at least one of between a square tree and a binary tree.
In the method and the device for encoding a video signal in accordance with the present invention, the determination of the intra prediction mode may include: generating a list of candidates having multiple candidates; and determine the intra prediction mode of the current block based on the list of candidates and an index.
In the procedure and the device for encoding a video signal in accordance with the present invention, a maximum number of candidates that may be included in the list of candidates may be more than three.
In the method and the device for encoding a video signal in accordance with the present invention, the determined intra prediction mode may be one of the intra extended prediction modes, and the extended intra prediction modes include a flat mode, a mode DC and more than 33 directional prediction modes.
In the method and the device for encoding a video signal in accordance with the present invention, the realization of the intra prediction may include: obtaining a prediction sample of the current block based on the intra prediction mode and the reference sample; and correct the prediction sample obtained using the differential information of the neighboring samples of the current block.
In the method and the device for encoding a video signal in accordance with the present invention, performing the intra prediction may include: obtaining a prediction sample of the current block based on the intra prediction mode and the reference sample; and correct the prediction sample obtained based on at least one of the predetermined weight and displacement.
Invention mode
5
10
fifteen
twenty
25
30
35
Various modifications can be made to the present invention and there are various embodiments of the present invention, the examples of which will now be provided with reference to the drawings and will be described in detail. Similar reference numbers refer to the similar element in the described drawings.
The terms used in the specification, "first", "second", etc. they can be used to describe various components, but the components should not be construed as limited to the terms. The terms are only used to differentiate one component from other components. For example, the "first" component may be called the "second" component without departing from the scope of the present invention, and the "second" component may also be similarly called the "first" component. The term "and / or" includes a combination of a plurality of articles or any one of a plurality of terms.
It will be understood that when a (first) element is simply referred to as being 'connected to' or 'coupled to' another (second) element without being 'directly connected to' or 'directly coupled to' this other (second) element In the present description, the (first) element may be 'directly' connected to 'or' directly coupled to 'a (third) intermediate element interposed between the two elements referred to above (first and second). On the contrary, it should be understood that when an element is referred to as being "directly coupled to" or "directly connected to" another element, no intermediate elements are present.
In the present specification, it should be understood that expressions such as "including", "having", etc. They are intended to indicate the existence of the characteristics, numbers, stages, actions, elements, parts or combinations thereof disclosed in the specification, and are not intended to exclude the possibility that one or more characteristics, numbers or numbers may be added. , stages, actions, elements, parts or combinations thereof.
Hereinafter, the preferred embodiments of the present invention will be described in detail with reference to the accompanying drawings. Hereinafter, the same constituent elements in the drawings are indicated with the same reference numbers, and a repeated description of the same elements will be omitted.
5
10
fifteen
twenty
25
30
35
Figure 1 is a block diagram illustrating a device for encoding a video in accordance with an embodiment of the present invention.
Referring to FIG. 1, the device 100 for encoding a video may include: an image partitioning module 110, prediction modules 120 and 125, a transformation module 130, a quantization module 135, a rearrangement module 160 , an entropy coding module 165, a reverse quantization module 140, a reverse transformation module 145, a filter module 150, and a memory 155.
The constituent parts shown in Figure 1 are shown independently to represent characteristic functions different from each other in the device for encoding a video. Therefore, it does not mean that each constituent part is constituted in a separate unit of hardware or software. Therefore, at least two constituent parts of each constituent part can be combined to form a constitutive part or a constitutive part can be partitioned into a plurality of constituent parts to perform each function. The embodiment where each constituent part is combined and the embodiment where a constitutive part is divided are also included in the scope of the present invention, if they do not depart from the essence of the present invention.
In addition, some of the constituents may not be indispensable constituents that perform essential functions of the present invention, but are selective constituents that improve only the performance thereof. The present invention can be implemented including only the constituent parts indispensable to implement the essence of the present invention, except the constituents used to improve performance. The structure that includes only the essential constituents except the selective constituents used to improve performance alone is also included within the scope of the present invention.
The image partitioning module 110 may partition an input image into one or more processing units. In this case, the processing unit may be a prediction unit (PU), a transformation unit (TU), or a coding unit (CU). PU, TU and CU are concepts specified in the state of the art: "Overview of the High Efficiency Video Coding (HEVC) Standard" IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, page 1651, Vol. 22, No. 12,
10
5
10
fifteen
twenty
25
30
35
December 2012. Image partitioning module 110 can partition an image into combinations of multiple encoding units, prediction units, and transformation units, and can encode an image by selecting a combination of encoding units, prediction units, and units of transformation with a predetermined criterion (for example, the cost function).
For example, an image can be partitioned into multiple encoding units. A recursive tree structure, such as a square tree structure, can be used to partition an image into coding units. An encoding unit that is partitioned into other encoding units with a larger image or encoding unit as a root can be partitioned with secondary nodes corresponding to the number of partitioned encoding units. An encoding unit that is not partitioned by a predetermined limitation serves as a leaf node. That is, when it is assumed that only square partitioning is possible for one coding unit, one coding unit can be partitioned into four other coding units at most.
Hereinafter, in the embodiment of the present invention, the coding unit may mean a unit that performs the coding, or a unit that performs the decoding.
A single coding unit can be partitioned into at least one prediction unit that has a square or rectangular shape. The prediction units resulting from the partition of a single coding unit may have the same size, or a prediction unit resulting from the partition (partitioning) of a single coding unit may have a different shape and / or size than another prediction unit partitioned in the same coding unit.
When a prediction unit undergoes an intra prediction it is generated based on a coding unit and the coding unit is not the smallest coding unit, the intra prediction can be performed without partitioning the prediction unit into multiple prediction units NxN .
The prediction modules 120 and 125 may include an inter prediction module 120 that performs an inter prediction and an intra prediction module 125 that performs an intra prediction. If it can be determined to make an inter prediction or an intra prediction for the
eleven
5
10
fifteen
twenty
25
30
35
prediction, the detailed information (e.g., an intra prediction mode, a motion vector, a reference image, etc.) can be determined according to each prediction procedure. In this case, the predicted processing unit may be different from the processing unit for which the prediction procedure and the detailed content are determined. For example, the prediction procedure, the prediction mode, etc. they can be determined by prediction unit, and the prediction can be made by the transformation unit. A residual value (residual block) can be entered between the generated prediction block and an original block to the transformation module 130. In addition, the prediction mode information used for the prediction, the motion vector information, etc. it can be encoded with the residual value by the entropy coding module 165 and can be transmitted to a device to decode a video. When a specific coding mode is used, the original block may be intactly encoded and transmitted to a decoding module without generating the prediction block through the prediction module 120 and 125.
The inter prediction module 120 can predict the prediction unit based on the information of at least one of a previous image or a subsequent image of the current image, or it can predict the prediction unit based on the information of some regions encoded in the current image, in some cases. The inter prediction module 120 may include an interpolation module of reference images, a motion prediction module, and a motion compensation module.
The reference image interpolation module can receive reference image information from memory 155 and can generate information of an entire pixel or less of the reference image. In the case of luma pixels, an interpolation filter based on DCT of 8 sockets having different filter coefficients can be used to generate the pixel information of an entire pixel or less in units of a 1/4 pixel. In the case of chroma signals, a 4-socket DCT based interpolation filter that has different filter coefficient can be used to generate the pixel information of an entire pixel or less in units of a 1/8 pixel.
The motion prediction module can perform the motion prediction based on the reference image interpolated by the reference image interpolation module. As procedures for calculating a motion vector, various procedures can be used, such as a block comparison algorithm based on
12
5
10
fifteen
twenty
25
30
35
full search (FBMA), a three-stage search algorithm (TSS), a new three-stage search algorithm (NTS), etc. The motion vector can have a motion vector value in units of 1/2 of a pixel or a 1/4 of a pixel based on an interpolated pixel. The motion prediction module can predict a current prediction unit by changing the motion prediction procedure. Various procedures, such as a skip procedure, a mixing procedure, an AMVP (advanced motion vector prediction) procedure, an intra block copy procedure, etc., can be used as motion prediction procedures.
The intra prediction module 125 can generate a prediction unit based on the reference pixel information next to a current block that is the pixel information in the current image. When the neighboring block of the current prediction unit is a block under inter prediction and, therefore, a reference pixel is a pixel under inter prediction, the reference pixel included in the block under inter prediction can be used. instead of the reference pixel information of a neighboring block subjected to intra prediction. That is, when a reference pixel is not available, at least one other reference pixel can be used than if it is available instead of the reference pixel information not available.
The prediction modes in the intra prediction may include a directional prediction mode using the reference pixel information as a function of a prediction direction and a non-directional prediction mode not using the directional information in the prediction realization. One way to predict luma information may be different from one way to predict chroma information, and in order to predict chroma information, intra prediction mode information used to predict luma information or signal information may be used. Luma predicted.
When making the intra prediction, when the size of the prediction unit is the same as the size of the transformation unit, the intra prediction can be made in the prediction unit based on the pixels located on the left, in the upper left , and at the top of the prediction unit. However, when performing the intra prediction, when the prediction unit size is different from the transformation unit size, the intra prediction can be made using a reference pixel based on the transformation unit. In addition, intra prediction can be used using NxN partitioning only for the smallest codification unit.
13
5
10
fifteen
twenty
25
30
35
In the intra prediction procedure, a prediction block can be generated after applying an AIS filter (intra adaptive smoothing) to a reference pixel depending on the prediction modes. The type of the AIS filter applied to the reference pixel may vary. In order to perform the intra prediction procedure, an intra prediction mode of the current prediction unit can be predicted from the intra prediction mode of the neighboring prediction unit to the current prediction unit. In predicting the prediction mode of the current prediction unit using the predicted mode information from the neighboring prediction unit, when the intra prediction mode of the current prediction unit is the same as the intra prediction mode of the neighboring prediction unit, the information indicating that the prediction modes of the current prediction unit and the neighboring prediction unit are equal to each other can be transmitted using the predetermined indication information. When the prediction mode of the current prediction unit is different from the prediction mode of the neighboring prediction unit, entropy coding can be performed to encode the prediction mode information of the current block.
In addition, a residual block that includes information on a residual value that is different between the prediction unit and the original block can be generated based on the prediction units generated by the prediction modules 120 and 125. The generated residual block can be introduced in transformation module 130.
The transformation module 130 can transform the residual block that includes information on the residual value between the original block and the prediction unit generated by the prediction modules 120 and 125 using a transformation procedure, such as the discrete cosine transform ( DCT), the discrete sinus transform (DST) and the KLT. Although applying DCT, DST or KLT in order to transform the residual block can be determined based on the intra prediction mode information of the prediction unit used to generate the residual block.
The quantization module 135 can quantify the values transmitted to a frequency domain by the transformation module 130. Quantification coefficients may vary depending on the block or the importance of an image. The values calculated by the quantization module 135 can be provided to the inverse quantization module 140 and the rearrangement module 160.
5
10
fifteen
twenty
25
30
35
The rearrangement module 160 can reorder the coefficients of the quantified residual values.
The rearrangement module 160 can change a coefficient in the form of a two-dimensional block into a coefficient in the form of a one-dimensional vector through the coefficient scanning procedure. For example, the rearrangement module 160 can explore from a DC coefficient to a coefficient in a high frequency domain using a zigzag scanning procedure in order to change the coefficients to be in the form of one-dimensional vectors. Depending on the size of the transformation unit and the intra prediction mode, the vertical direction scan where the coefficients in the form of two-dimensional blocks are explored in the direction of the column or, can be used instead of the zigzag scan. the horizontal direction scan where the coefficients in the form of two-dimensional blocks are explored in the direction of the row. That is, it can be determined which scanning procedure is used between the zigzag scan, the vertical direction scan and the horizontal direction scan depending on the size of the transformation unit and the intra prediction mode.
The entropy coding module 165 can perform the entropy codification based on the values calculated by the rearrangement module 160. Entropy coding can use various coding procedures, for example, exponential Golomb coding, context-adaptable variable length coding (CAVLC), and context-adaptable binary arithmetic coding (CABAC).
Entropy coding module 165 may encode various information, such as residual value coefficient information and block type information of the coding unit, prediction mode information, partition unit information, information prediction unit, transformation unit information, motion vector information, reference frame information, block interpolation information, filtering information, etc. of the rearrangement module 160 and the prediction modules 120 and 125.
The entropy coding module 165 can encode the entropy of the input coefficients of the coding unit from the rearrangement module 160.
The inverse quantization module 140 can inversely quantify the values
fifteen
5
10
fifteen
twenty
25
30
35
quantified by the quantization module 135 and the inverse transformation module 145 can reverse transform the values transformed by the transformation module 130. The residual value generated by the inverse quantization module 140 and the inverse transformation module 145 can be combined with the prediction unit predicted by a motion estimation module, a motion compensation unit, and the intra prediction module of the modules 120 and 125 prediction such that a rebuilt block can be generated.
The filter module 150 may include at least one of an unblocking filter, a displacement correction unit, and an adaptive loop filter (ALF).
The unlock filter can eliminate block distortion that occurs due to the boundaries between the blocks in the reconstructed image. In order to determine if the unlocking should be performed, it can be determined whether to apply the unblocking filter to the current block based on the pixels included in several rows or columns in the block. When the unblocking filter is applied to the block, a strong filter or a weak filter can be applied depending on the required unblocking filtering force. Furthermore, when applying the unblocking filter, the filtering in the horizontal direction and the filtering in the vertical direction can be processed in parallel.
The displacement correction module can correct the displacement of the original image with respect to the image subject to unlocking in units of one pixel. In order to perform the offset correction in a specific image, it is possible to use a method of applying a shift in consideration of the edge information of each pixel or a pixel partitioning procedure of an image in the predetermined number of regions, determining a region to undergo the displacement and apply the displacement to the given region.
Adaptive loop filtering (ALF) can be performed based on the value obtained by comparing the reconstructed filtered image and the original image. The pixels included in the image can be partitioned into predetermined groups, a filter to be applied to each of the groups can be determined, and the filtering can be performed individually for each group. Information on whether to apply ALF and a luma signal can be transmitted by the codification units (CU). The shape and filter coefficient of a filter for ALF may vary depending on each block. In addition, the filter for ALF in the same form (fixed form) can be applied regardless of the characteristics of the target block of
16
5
10
fifteen
twenty
25
30
35
Memory 155 can store the reconstructed block or image calculated through filter module 150. The reconstructed block or image stored can be provided to prediction modules 120 and 125 when performing inter prediction.
Figure 2 is a block diagram illustrating a device for decoding a video in accordance with an embodiment of the present invention.
Referring to FIG. 2, the device 200 for decoding a video may include: an entropy decoding module 210, a rearrangement module 215, a reverse quantization module 220, a reverse transformation module 225, modules 230 and 235 prediction, a filter module 240 and a memory 245.
When a video bit stream is input from the device to encode a video, the input bit stream can be decoded according to a reverse procedure of the device to encode a video.
The entropy decoding module 210 can perform the entropy decoding according to a reverse entropy coding procedure by the entropy coding module of the device for encoding a video. For example, corresponding to the procedures performed by the device for encoding a video, various procedures can be applied, such as exponential Golomb coding, context-adaptable variable length coding (CAVLC), and context-adaptable binary arithmetic coding (CABAC). ).
The entropy decoding module 210 can decode the information about the intra prediction and the inter prediction made by the device to encode a video.
The reordering module 215 can perform the reordering of the bitstream entropy decoded by the entropy decoding module 210 based on the reordering procedure used in the device to encode a video. The rearrangement module can reconstruct and reorganize the coefficients in the form of one-dimensional vectors for the coefficient in the form of two two-dimensional blocks. The reordering module 215 can reorder upon receiving the information related to the coefficient scan performed on the device to encode a
17
5
10
fifteen
twenty
25
30
35
video and inversely explore the coefficients based on the order of exploration performed on the device to encode a video.
The inverse quantization module 220 can perform the inverse quantization based on a quantification parameter received from the device to encode a video and the reordered coefficients of the block.
The inverse transformation module 225 can perform the inverse transformation, that is, the inverse DCT, the inverse DST, and the inverse KLT, which is the inverse of the transformation, that is, the DCT, the DST, and the KLT, performed by the transformation module on the quantification result by the device to encode a video. The reverse transformation can be performed based on the transfer unit determined by the device to encode a video. The reverse transformation module 225 of the device for decoding a video can selectively perform transformation techniques (for example, DCT, DST, and KLT) based on multiple pieces of information, such as the prediction procedure, the size of the current block, the prediction direction, etc.
The prediction modules 230 and 235 can generate a prediction block based on the information about the generation of the prediction block received from the entropy decoding module 210 and the previously decoded block or image information received from the memory 245.
As described above, such as the operation of the device for encoding a video, in the realization of intra prediction, when the size of the prediction unit is the same as the size of the transformation unit, the intra prediction can be performed in the prediction unit based on the pixels located to the left, in the upper left, and in the upper part of the prediction unit. When performing the intra prediction, when the size of the prediction unit is different from the size of the transformation unit, the intra prediction can be made using a reference pixel based on the transformation unit. In addition, intra prediction can be used using NxN partitioning only for the smallest codification unit.
The prediction modules 230 and 235 may include a prediction unit determination module, an inter prediction module, and an intra prediction module. The prediction unit determination module can receive various information, such as the prediction unit information, the prediction mode information of a
18
5
10
fifteen
twenty
25
30
35
intra prediction procedure, information on the prediction of movement of an inter prediction procedure, etc. From module 210 of entropy decoding, you can distinguish a unit prediction in a current codification unit, and you can determine whether the inter prediction or the intra prediction is made in the prediction unit. Using the necessary information in the inter prediction of the current prediction unit received from the device to encode a video, the inter prediction module 230 can make the inter prediction in the current prediction unit based on the information of at least one of the a previous image or a subsequent image of the current image that includes the current prediction unit. As an alternative, inter prediction can be made based on the information of some pre-reconstructed regions in the current image that includes the current prediction unit.
In order to make the inter prediction, it can be determined which of a skip mode, a mix mode, an AMVP mode, and an inter block copy mode is used as the motion prediction procedure of the unit of prediction included in the coding unit based on the coding unit.
The intra 235 prediction module can generate a prediction block based on the pixel information in the current image. When the prediction unit is a prediction unit subject to intra prediction, intra prediction can be performed based on the intra prediction mode information of the prediction unit received from the device for encoding a video. The intra 235 prediction module may include an intra-adaptive smoothing filter (AIS), a reference pixel interpolation module, and a DC filter. The AIS filter filters the reference pixel of the current block, and the possibility of applying the filter can be determined according to the prediction mode of the current prediction unit. AIS filtering can be performed on the reference pixel of the current block using the prediction mode of the prediction unit received from the device to encode a video and the AIS filter information. When the prediction mode of the current block is a mode where AIS filtering is not performed, the AIS filter cannot be applied.
When the prediction unit prediction mode is a prediction mode in which the intra prediction is made based on the pixel value obtained by interpolating the reference pixel, the reference pixel interpolation module can interpolate the reference pixel to generate a pixel of an entire pixel or less. When the prediction mode of the current prediction unit is a prediction mode in which it is generated
19
5
10
fifteen
twenty
25
30
35
A prediction block without interpolating the reference pixel, the reference pixel cannot be interpolated. The DC filter can generate a prediction block through filtering when the prediction mode of the current block is a DC mode.
The reconstructed block or image can be provided to filter module 240. The filter module 240 may include the unblocking filter, the offset correction module, and the ALF.
Information on whether or not the unblocking filter is applied to the corresponding block or image, and the information on which the strong filter and the weak filter is applied when the unblocking filter is applied can be received from the device to encode a video. The unlocking filter of the device for decoding a video can receive information about the unlocking filter from the device for encoding a video, and can perform an unblocking filtering on the corresponding block.
The shift correction module can make a shift correction in the reconstructed image based on the type of offset correction applied to an image when coding and the offset value information.
The ALF can be applied to the coding unit based on the information on whether to apply the ALF, the ALF coefficient information, etc. received from the device to encode a video. ALF information can be provided as being included in a specific set of parameters.
Memory 245 may store the image or the reconstructed block for use as an image or a reference block, and may provide the reconstructed image to an output module.
As described above, in the realization of the present invention, for the sake of explanation, the coding unit is used as a term representing a coding unit, but the coding unit can serve as a unit that performs decoding as well. as coding
Figure 3 is a view illustrating an example for hierarchically partitioning a coding block based on a tree structure according to an embodiment of the
twenty
5
10
fifteen
twenty
25
30
35
An input video signal is decoded into predetermined block units, and a predetermined unit for decoding the input video signal is a codification block. The codification block can be a unit on which intra / inter prediction, transformation, and quantification are performed. The coding block can be a square or non-square block that has an arbitrary size in a range of 8x8 to 64x64, or it can be a square or non-square block that has a size of 128x128, 256x256 or more.
Specifically, the coding block can be partitioned hierarchically based on at least one of a square tree and a binary tree. In this case, partitioning based on the square tree can mean that a 2Nx2N coding block is partitioned into four NxN coding blocks, and binary tree based partitioning can mean that one coding block is partitioned into two blocks of coding. Partitioning based on the binary tree can be done symmetrically or asymmetrically. The partitioned coding block based on the binary tree can be a square block or a non-square block, such as a rectangular shape. Partitioning based on the binary tree can be performed in a coding block where partitioning based on the square tree is no longer performed. Partitioning based on the square tree can no longer be performed on the partitioned coding block based on the binary tree.
In order to perform an adaptive partitioning based on the square tree or the binary tree, the information indicating the partitioning based on the square tree, the information on the size / depth of the coding block that the partitioning based on the tree can be used binary tree is allowed, the information indicating partitioning based on the binary tree, information about the size / depth of the coding block that partitioning based on the binary tree is allowed, information about the size / depth of the coding block that partitioning based on the binary tree is not allowed, information on whether partitioning based on the binary tree is performed in a vertical direction or in a horizontal direction, etc.
As shown in Figure 3, the first coding block 300 with the partition depth of k can be partitioned into multiple second blocks
twenty-one
5
10
fifteen
twenty
25
30
35
of coding based on the square tree. For example, the second coding blocks 310 to 340 can be square blocks that are half the width and half the height of the first coding block, and the partition depth of the second coding block can be increased to k + 1.
The second coding block 310 with the partition depth of k +1 can be partitioned into multiple third coding blocks with the partition depth of k + 2. Partitioning of the second coding block 310 can be performed using one of the tree selectively square and the binary tree depending on a partition mode. In this case, partitioning can be determined based on at least one of the information indicating partitioning based on a square tree and information indicating partitioning based on a binary tree.
When the second coding block 310 is partitioned based on the square tree, the second coding block 310 can be partitioned into four third coding blocks 310a that are half the width and half the height of the second coding block, and the depth The partition of the third coding block 310a can be increased by k + 2. In contrast, when the second coding block 310 is partitioned based on the binary tree, the second coding block 310 can be partitioned into two third coding blocks. In this case, each of the two third coding blocks can be a non-square block that is one half the width and half the height of the second coding block, and the partition depth can be increased by k + 2. The second coding block can be determined as a non-square block of a horizontal direction or a vertical direction depending on a partition address, and the partition address can be determined based on the information about whether partitioning based on the binary tree is performed in a vertical direction or in a horizontal direction.
Meanwhile, the second coding block 310 can be determined as a sheet coding block that is no longer partitioned based on the square tree or the binary tree. In this case, the sheet coding block can be used as a prediction block or a transformation block.
Like partitioning of the second coding block 310, the third coding block 310a can be determined as a sheet coding block, or it can be further partitioned based on the square tree or the binary tree.
22
5
10
fifteen
twenty
25
30
35
Meanwhile, the third partitioned coding block 310b based on the binary tree can also be partitioned into the coding blocks 310b-2 of a vertical direction or the coding blocks 310b-3 of a horizontal direction based on the binary tree, and The partition depth of the relevant coding blocks can be increased by k + 3. Alternatively, the third coding block 310b can be determined as a sheet coding block 310b-1 that is no longer partitioned based on the binary tree. In this case, the coding block 310b-1 can be used as a prediction block or a transformation block. However, the above partitioning procedure can be performed in a limited manner based on at least one of the information on the size / depth of the coding block that partitioning based on the square tree is allowed, the information on the size / depth of the coding block that partitioning based on the binary tree is allowed, and information on the size / depth of the coding block partitioning based on the binary tree is not allowed.
Figure 4 is a view illustrating the types of predefined intra prediction modes for a device for encoding / decoding a video in accordance with an embodiment of the present invention.
The device for encoding / decoding a video can perform intra prediction using one of the predefined intra prediction modes. The predefined intra prediction modes for intra prediction can include non-directional prediction modes (eg, a flat mode, a DC mode) and 33 directional prediction modes.
Alternatively, in order to improve the accuracy of intra prediction, a large number of directional prediction modes than the 33 directional prediction modes can be used. That is, M extended directional prediction modes can be defined by subdividing the angles of the directional prediction modes (M> 33), and a directional prediction mode having a predetermined angle can be obtained using at least one of the 33 directional prediction modes predefined.
Figure 4 shows an example of intra-extended prediction modes, and intra-extended prediction modes can include two non-directional prediction modes and 65 extended directional prediction modes. The same numbers of intra-extended prediction modes can be used for a luma component and a component
2. 3
5
10
fifteen
twenty
25
30
35
chroma, or a different number of intra prediction modes can be used for each component. For example, 67 intra-extended prediction modes can be used for the luma component, and 35 intra prediction modes can be used for the chroma component.
Alternatively, depending on the chroma format, a different number of intra prediction modes can be used in the realization of the intra prediction. For example, in the case of the 4: 2: 0 format, 67 intra prediction modes can be used for the luma component to perform the intra prediction and 35 intra prediction modes can be used for the chroma component. In the case of the 4: 4: 4 format, 67 intra prediction modes can be used for both the luma component and the chroma component to perform the intra prediction.
Alternatively, depending on the size and / or shape of the block, a different number of intra prediction modes can be used to perform the intra prediction. That is, depending on the size and / or the shape of the PU or CU, 35 intra prediction modes or 67 intra prediction modes can be used to perform the intra prediction. For example, when the CU or PU has a size of less than 64x64 or is partitioned asymmetrically, 35 of intra prediction modes can be used to perform intra prediction. When the size of the CU or PU is equal to or greater than 64x64, 67 intra prediction modes can be used to perform the intra prediction. 65 intra-directional prediction modes can be allowed for Intra_2Nx2N, and only 35 intra-directional prediction modes can be allowed for Intra_NxN.
Fig. 5 is a flow chart that briefly illustrates an intra prediction procedure according to an embodiment of the present invention.
Referring to Figure 5, an intra prediction mode of the current block can be determined in step S500.
Specifically, the intra prediction mode of the current block can be obtained based on a list of candidates and an index. In this case, the candidate list contains multiple candidates, and multiple candidates can be determined based on an intra prediction mode of the neighboring block adjacent to the current block. The neighboring block may include at least one of the blocks located at the top, bottom, left, right and corner of the current block. The index can specify one of the multiple candidates from the list of candidates. The candidate specified by the index can
24
5
10
fifteen
twenty
25
30
35
set to the intra prediction mode of the current block.
An intra prediction mode used for intra prediction in the neighboring block can be established as a candidate. In addition, an intra prediction mode that has a directionality similar to that of the intra prediction mode of the neighboring block can be established as a candidate. In this case, the intra prediction mode that has similar directionality can be determined by adding or subtracting a predetermined constant value to or from the intra prediction mode of the neighboring block. The predetermined constant value can be an integer, such as one, two or more.
The list of candidates may include a default mode. The default mode may include at least one of a flat mode, a DC mode, a vertical mode, and a horizontal mode. The default mode can be added adaptively taking into account the maximum number of candidates that can be included in the list of candidates in the current block.
The maximum number of candidates that can be included in the list of candidates can be three, four, five, six, or more. The maximum number of candidates that can be included in the list of candidates can be a fixed preset value in the device to encode / decode a video, or it can be determined differently based on a characteristic of the current block. The characteristic can mean the
block location / size / shape, the number / type of intra prediction modes that the block can use, etc. Alternatively, the information indicating the maximum number of candidates that can be included in the list of candidates can be indicated separately, and the maximum number of candidates that can be included in the list of candidates can be determined differently using the information. The information indicating the maximum number of candidates can be indicated in at least one of a sequence level, an image level, a segment level and a block level.
When the intra-extended prediction modes and the 35 predefined intra-prediction modes are used selectively, the intra prediction modes of the neighboring blocks can be converted into indices corresponding to the extended intra prediction modes, or the indices corresponding to the 35 Intra prediction modes, with which candidates can be obtained. To transform a index, a predefined table can be used, or a scaling operation can be used based on a predetermined value. In this case, the predefined table can define a mapping relationship
25
5
10
fifteen
twenty
25
between different groups of intra prediction modes (for example, extended intra prediction modes and 35 intra prediction modes).
For example, when the left neighboring block uses the 35 intra prediction modes and the intra prediction mode of the left neighboring block is 10 (a horizontal mode), it can be transformed into a 16-digit index corresponding to a horizontal mode in the prediction modes intra extended.
Alternatively, when the upper neighboring block uses the intra extended prediction modes and the intra prediction mode, the upper neighboring block has an index of 50 (a vertical mode), which can be transformed into an index of 26 corresponding to a vertical mode in the 35 intra prediction modes.
Based on the procedure described above to determine the intra prediction mode, the intra prediction mode can be obtained independently for each of the luma component and the chroma component, or the intra prediction mode of the chroma component can be obtained depending on the mode of intra prediction of the luma component.
Specifically, the intra prediction mode of the chroma component can be determined based on the intra prediction mode of the luma component as shown in the following Table 1.
[Table 1]
 Intra_chroma_pred_mode [xCb] [yCb]  IntraPredModeY [xCb] [yCb]
 0  26 10 1 X (0 <= X <= 34)
 0  3. 4  0 0 0 0
 one  26 34 26 26 26
 2  10 10 34 10 10
 3  1 1 1 34 1
 4  0 26 10 1 X
5
10
fifteen
twenty
25
30
35
In Table 1, intra_chroma_pred_mode means the information indicated to specify the intra prediction mode of the chroma component, and IntraPredModeY indicates the intra prediction mode of the luma component.
Referring to Figure 5, a reference sample can be obtained for the intra prediction of the current block in step S510.
Specifically, a reference sample for intra prediction can be obtained based on a neighboring sample of the current block. The neighboring sample can be a reconstructed sample of the neighboring block, and the reconstructed sample can be a reconstructed sample before a loop filter is applied or a reconstructed sample is applied after applying the loop filter.
A reconstructed neighboring sample before the current block can be used as the reference sample, and a filtered neighbor sample based on a predetermined intra filter can be used as the reference sample. The intra filter may include at least one of the first intra filter applied to multiple neighboring samples located in the same horizontal line and the second intra filter applied to multiple neighboring samples located in the same vertical line. Depending on the positions of the neighboring samples, one of the first intra filter and the second intra filter can be applied selectively, or both intra filters can be applied.
Filtering can be done adaptively based on at least one of the intra prediction modes of the current block and the size of the transformation block of the current block. For example, when the intra prediction mode of the current block is DC mode, it is possible that vertical mode or horizontal mode filtering is not performed. When the size of the transformation block is NxM, filtering may not be performed. In this case, N and M may be the same or different values, or they may be values of 4, 8, 16 or more. Alternatively, filtering can be performed selectively based on the result of a comparison of a predefined threshold and the difference between the intra prediction mode of the current block and the vertical mode (or the horizontal mode). For example, when the difference between the intra prediction mode of the current block and the vertical mode is greater than a threshold, filtering can be performed. The threshold can be defined for each size of the transformation block as shown in Table 2.
5
10
fifteen
twenty
25
30
[Table 2]
 8x8 Transformed 16x16 Transformed 32x32 Transformed
 Limit  7 1 0
The intra filter can be determined as one of the multiple predefined intra filter candidates in the device for encoding / decoding a video. To this end, a code can be indicated that specifies an intra filter of the current block among the multiple candidates of intra filter. Alternatively, the intra filter can be determined based on at least one of the size / shape of the current block, the size / shape of the transformation block, the information on the intensity of the filter and the variations of the neighboring samples.
Referring to Figure 5, the intra prediction can be performed using the intra prediction mode of the current block and the reference sample in step S520.
That is, the prediction sample of the current block can be obtained using the intra prediction mode determined in step S500 and the reference sample obtained in step S510. However, in the case of intra prediction, a sample at the edge of the neighboring block can be used, and therefore the quality of the prediction image can be decreased. Therefore, a correction procedure can be performed on the prediction sample generated through the prediction procedure described above, and which will be described in detail with reference to Figures 6 to 15. However, the correction procedure is not limited. to be applied only to the intra prediction sample, and can be applied to an inter prediction sample or to the reconstructed sample.
Figure 6 is a view illustrating a procedure for correcting a prediction sample of a current block based on the differential information of neighboring samples according to an embodiment of the present invention.
The prediction sample of the current block can be corrected based on the differential information of multiple neighboring samples for the current block. The correction can be made on all prediction samples in the current block, or it can be made on prediction samples in some predetermined regions. Some regions can be a row / column or several rows / columns, or they can be preset regions for your
5
10
fifteen
twenty
25
30
correction in the device to encode / decode a video, or they can be determined differently based on at least one of the size / shape of the current block and the intra prediction mode.
Neighboring samples may belong to neighboring blocks placed in the upper, left and upper left corner of the current block. The number of neighboring samples used for the correction can be two, three, four or more. The positions of the neighboring samples can be determined differently depending on the position of the prediction sample which is the correction objective in the current block. Alternatively, some of the neighboring samples may have fixed positions regardless of the position of the prediction sample that is the correction objective, and the remaining neighboring samples may have different positions determined as a function of the position of the prediction sample that is The objective of correction.
The differential information of the neighboring samples can mean a differential sample between the neighboring samples, or it can mean a value obtained by scaling the differential sample by a predetermined constant value (for example, one, two, three, etc.). In this case, the predetermined constant value can be determined by considering the position of the prediction sample that is the correction objective, the position of the column or row that includes the prediction sample that is the correction objective, the position of the sample of prediction within the column or row, etc.
For example, when the intra prediction mode of the current block is the vertical mode, differential samples between the upper left neighboring sample p (-1, -1) and the neighboring samples p (-1, y) adjacent to the limit can be used left of the current block to obtain the final prediction sample as shown in Formula 1. (y = 0 ... N-1)
[Formula 1]
P '(0, y) = P (0, y) + (p (-1, y) -) >> 1
In Formulas 1-6, the ">>" operator denotes the bitwise shift to the right.
5
10
fifteen
twenty
25
30
35
For example, when the intra-prediction mode of the current block is the horizontal mode, the differential samples between the upper left neighboring sample p (-1, -1) and the neighboring samples p (x, -1) adjacent to the Kmite can be used top of the current block to obtain the final prediction sample as shown in Formula 2. (x = 0 ... N-1)
[Formula 2]
P '(x, 0) = p (x, 0) + (p (x, -1) - p (-1, -1)) >> 1
For example, when the intra prediction mode of the current block is the vertical mode, differential samples between the upper left neighboring sample p (-1, -1) and the neighboring samples p (-1, y) adjacent to the limit can be used left of the current block to get the final prediction sample. In this case, the differential sample can be added to the prediction sample, or the differential sample can be scaled by a predetermined constant value, and then added to the prediction sample. The predetermined constant value used in the scale can be determined differently depending on the column and / or the row. For example, the prediction sample can be corrected as shown in Formula 3 and Formula 4. (y = 0 ... N-1)
[Formula 3]
P '(0, y) = P (0, y) + (p (-1, y) -p (-1, -1)) >> 1 [Formula 4]
P '(1, y) = P (1, y) + (p (-1, y) - p (-1, -1)) >> 2
For example, when the intra-prediction mode of the current block is the horizontal mode, differential samples between the upper left neighboring sample p (-1, -1) and neighboring samples p (x, -1) adjacent to the limit can be used top of the current block to obtain the final prediction sample, as described in the case of vertical mode. For example, the prediction sample can be corrected as shown in Formula 5 and Formula 6. (x = 0 .... N-1)
[Formula 5]
P '(x, 0) = p (x, 0) + (p (x, -1) -p (-1, -1)) >> 1 [Formula 6]
P '(x, 1) = p (x, 1) + (p (x, -1) -p (-1, -1)) >> 2
Figures 7 and 8 are views illustrating a procedure to correct a sample of
30
5
10
fifteen
twenty
25
30
35
prediction based on a predetermined correction filter according to an embodiment of the present invention.
The prediction sample can be corrected based on the neighboring sample of the prediction sample that is the correction objective and a predetermined correction filter. In this case, the neighboring sample may be specified by an angular line of the directional prediction mode of the current block, or it may be at least one sample placed on the same angular line as the prediction sample that is the correction target. In addition, the neighboring sample may be a prediction sample in the current block, or it may be a reconstructed sample in a reconstructed neighboring block before the current block.
At least one of the number of leads, the force, and a filter coefficient of the correction filter can be determined based on at least one of the position of the prediction sample that is the correction objective, whether the sample of prediction that is the correction objective is positioned or not within the limit of the current block, the intra prediction mode of the current block, the angle of the directional prediction mode, the prediction mode (the inter or intra mode) of the neighboring block, and the size / shape of the current block.
Referring to Figure 7, when the directional prediction mode has an index of 2 or 34, at least one prediction / reconstruction sample placed in the lower left of the prediction sample which is the correction target and the filter of Default correction can be used to obtain the final prediction sample. In this case, the prediction / reconstructed sample in the lower left can belong to a previous line of a line that includes the prediction sample that is the correction target, for the same block as the current sample, or for the block neighbor adjacent to the current block.
Filtering for the prediction sample can be done only in the line located on the block boundary, or it can be done in multiple lines. The correction filter can be used where at least one of the number of filter leads and a filter coefficient is different for each of the lines. For example, a filter (1/2, 1/2) can be used for the first left line closest to the block boundary, a filter (12/16, 4/16) can be used for the second line, a filter can be used (14/16, 2/16) for the third line, and a filter (15/16, 1/16) for the fourth line can be used.
31
5
10
fifteen
twenty
25
30
35
Alternatively, when the directional prediction mode has a rate of 3 to 6 or 30 to 33, filtering can be performed at the block limit as shown in Figure 8, and a 3-tap correction filter can be used to Correct the prediction sample. Filtering can be done using the lower left sample of the prediction sample that is the correction target, the lower sample of the lower left sample and a 3-shot correction filter that takes as input the prediction sample that is the target of correction. The position of the neighboring sample used by the correction filter can be determined differently based on the directional prediction mode. The filter coefficient of the correction filter can be determined differently depending on the directional prediction mode.
Different correction filters can be applied depending on whether the neighboring block is encoded in inter mode or in intra mode. When the neighboring block is encoded in the intra mode, a filtering procedure can be used in which more weight is given to the prediction sample, compared to when the neighboring block is encoded in the inter mode. For example, in the case that the intra prediction mode is 34, when the neighboring block is encoded in the inter mode, a filter (1/2, 1/2) can be used, and when the neighboring block is encoded in the intra mode, a filter (4/16, 12/16) can be used.
The number of lines to filter in the current block may vary depending on the size / shape of the current block (for example, the codification block and the prediction block). For example, when the size of the current block is equal to or less than 32x32, the filtering can be performed only on a line in the block boundary; otherwise, the filtering can be done in several lines, including the line in the block boundary.
Figures 7 and 8 are based on the case where the 35 intra prediction modes are used in the figure. 4, but can be applied equally / similarly to the case where intra-extended prediction modes are used.
Figure 9 is a view illustrating a procedure for correcting a prediction sample using weight and displacement according to an embodiment of the present invention.
When brightness changes between the previous frame and the current frame occurs, although the current block is similar to an attached block of the previous frame, than the image of
32
5
10
fifteen
twenty
25
30
35
prediction cannot be encoded in intra prediction or in inter prediction, or the prediction image quality encoded in intra prediction or in inter prediction can be relatively low. In this case, the weight and offset for brightness compensation can be applied to the prediction sample so that the quality of the prediction image can be improved.
Referring to Figure 9, at least one of the weight w and the displacement f can be determined in step S900.
At least one of the weight w and offset f can be indicated in at least one of a set of sequence parameters, a set of image parameters, and a segment header. Alternatively, at least one of the weight w and the displacement f can be indicated in predetermined block units sharing the same, and multiple blocks (e.g., CU, PU, and TU) belonging to a predetermined block unit that can share one of the weight w and / or the displacement f indicated.
At least one of the weight w and the displacement f can be indicated independently of the prediction mode of the current block, and can be selectively indicated considering the prediction mode. For example, when the prediction mode of the current block is the inter mode, the weight w and / or the displacement f can be indicated; otherwise, they cannot be indicated. In this case, the inter mode can include at least one of the skip mode, the mix mode, the AMVP mode and the current image reference mode. The current image reference mode can mean a prediction mode that uses a pre-reconstructed region in the current image that includes the current block. A motion vector can be used for the current image reference mode to specify the pre-reconstructed region. An indicator or index indicating whether the current block is encoded in the current image reference mode can be indicated, or can be obtained through a reference image index of the current block. The current image for the current image reference mode may exist in a fixed position (for example, the position with refIdx = 0 or the last position) in the reference image list of the current block. Alternatively, the current image may be placed differently in the reference image list, and for this purpose, a separate reference image index indicating the position of the current image may be indicated.
The weight can be obtained using the change of brightness between the first template, which has a
33
5
10
fifteen
twenty
25
30
certain template form and that is adjacent to the current block, and the second template that is adjacent to the block that goes before the current one. The second template may include a sample not available. In this case, an available sample can be copied to the position of the unavailable sample, or the available sample can be obtained by interpolation between multiple available samples. In this case, the available sample can be included in the second template or in the neighboring block. At least one of the coefficient, the shape and the number of shots of the filter used in the interpolation can be determined differently based on the size and / or shape of the template. A template composition procedure will be described in detail with reference to Figures 10 to 15.
For example, when the neighboring sample of the current block is designated by yi (i oscillating from 0 to N-1) and the neighboring sample from the attached block is designated by Xi (i oscillating from 0 to N-1), the weight w and the displacement f can be obtained as follows.
Using a template specifically, adjacent to the current block, weight w and displacement f can be obtained by obtaining the minimum value of E (w, f) in Formula 7.
[Formula 7]
image 1
Formula 7 to obtain the minimum value can be changed to Formula 8. [Formula 8]
image2
Formula 9 to obtain weight w and formula 10 to obtain displacement f can be obtained from formula 8.
[Formula 9]
] Y Yes Zi X] Yes yt I X
W “N £ f x, yxf - £, Si x; + X
[Formula 10]
5
10
fifteen
twenty
25
30
35
image3
Referring to Figure 9, at least one of the weight and displacement determined in step S900 can be used to correct the prediction sample.
For example, when a change in brightness occurs in all frames, the weight w and offset f are applied to the prediction sample p generated through the intra prediction such that a corrected prediction sample p 'can be obtained as It is shown in Formula 11.
[Formula 11]
p ’= w X p + f
In this case, the weight w and the displacement f can be applied to the prediction sample generated through the inter prediction, or they can be applied to the reconstructed sample.
Figures 10 to 15 are views illustrating a method of composing a template for determining the weight w in accordance with an embodiment of the present invention.
Referring to the left of Figure 10, a template may be composed of all neighboring samples adjacent to the current block, or a template may be composed of some subsampled samples from neighboring samples adjacent to the current block. The medium in Figure 10 shows an example of 1/2 subsampling, and a template can be composed only of gray samples. Instead of 1/2 subsampling, the template can be composed using 1/4 subsampling or 1/8 subsampling. As shown to the right of Figure 10, a template can be composed of all neighboring samples adjacent to the current block, except for the sample placed in the upper left. Not shown in Figure 10, considering the position of the current block in the image or a block of coding tree (the largest coding unit), a template composed only of the samples placed on the left or a composite template can be used only for the samples placed on top.
Referring to Figure 11, the template can be composed by increasing the number of neighboring samples. That is, the template in Figure 11 may be composed of the
5
10
fifteen
twenty
25
30
35
first neighboring samples adjacent to the Kmite of the current block and the second neighboring samples adjacent to the first neighboring samples.
As shown on the left of the figure. 11, a template can be composed of all neighboring samples belonging to two lines adjacent to the limit of the current block, or as shown in the middle of Figure 11, a template can be composed of a subsample of the template on the left . As shown to the right of Figure 11, a template can be composed excluding four samples belonging to the upper left. Not shown in Figure 11, considering the position of the current block in the image or a block of coding tree (the largest coding unit), a template composed solely of the samples placed on the left or a template composed only of one can be used for the samples placed on top.
Alternatively, different templates can be composed depending on the size and / or shape of the current block (if the current block has a square shape, if the current block is partitioned symmetrically). For example, as shown in Figure 12, a template subsampling rate can be applied differently depending on the size of the current block. For example, as shown on the left of Figure 12, when the block size is equal to or less than 64x64, 1/2 of subsample template can be composed. As shown to the right of Figure 12, when the block size is equal to or greater than 128x128, 1/4 of subsample template.
Referring to Figure 13, the template can be composed by increasing the number of neighboring samples adjacent to the current block depending on the size of the same.
Multiple template candidates that can be determined in a sequence or a slice can be used, and one of the multiple template candidates can be selectively used. Multiple template candidates can be templates of different shapes and / or sizes. Information on the shape and / or size of the template can be indicated in a sequence header or segment header. In the device for encoding / decoding a video, a template can be assigned to each template. In order to identify the template candidates to be used in the current sequence, image or segment among the multiple template candidates, the type_weight_pred_template_idx syntax can be encoded. The device for decoding a video can use template candidates selectively based on the type_weight_pred_template_idx syntax.
36
For example, as shown in Figure 14, the template in the middle of Figure 10 can be assigned to 0, the template in the right of Figure 10 can be assigned to 1, the template in the center of Figure 11 can be assigned to 2 , and the template on the right of Figure 11 can be assigned to 3. The template used in the sequence can be indicated.
When making the weighted prediction using a non-square block, the template can be composed by applying different subsampling rates for long and short sides so that the total number of templates is 2AN. For example, as shown in Figure 10 15, the template can be composed by performing 1/2 of subsampling on the short side and 1/4 of
subsampling on the long side
Industrial applicability
The present invention can be used to encode / decode a video signal.
权利要求:
Claims (15)
[1]
5
10
fifteen
twenty
25
30
35
1. A procedure for decoding a video, the procedure comprising: determining an intra prediction mode of a current block;
obtain a reference sample for an intra prediction of the current block; Y
make an intra prediction of the current block based on the intra prediction mode and
the reference sample,
wherein the current block is a partitioned codification block based on at least one of a square tree or a binary tree, and is a non-square block.
[2]
2. The method of claim 1, wherein the determination of the intra prediction mode comprises:
generate a list of candidates that includes multiple candidates; Y
determine the intra prediction mode of the current block based on the candidate list and index information, the index information specifying a candidate index included in the candidate list,
in which a maximum number of candidates that can be included in the list of candidates is more than three,
the intra-determined prediction mode is one of some extended intra prediction modes, and
Intra extended prediction modes include a flat mode, a DC mode and more than 33 directional prediction modes.
[3]
3. The method of claim 1, wherein the realization of the intra prediction comprises:
obtain a prediction sample of the current block based on the intra prediction mode and the reference sample; Y
correct the prediction sample obtained using the differential information of the neighboring samples of the current block.
[4]
4. The method of claim 1, wherein performing the intra prediction comprises:
obtain a prediction sample of the current block based on the intra prediction mode and the reference sample; Y
correct the prediction sample obtained based on at least one of a predetermined weight and a predetermined displacement.
38
5
10
fifteen
twenty
25
30
35
[5]
5. A device for decoding a video, the device comprising:
a prediction module that determines an intra prediction mode of a current block, that obtains a reference sample for the intra prediction of the current block and that makes an intra prediction of the current block based on the intra prediction mode and the sample of reference,
wherein the current block is a partitioned codification block based on at least one of a square tree or a binary tree, and is a non-square block.
[6]
6. The device of claim 5, wherein the prediction module generates a list of candidates that includes multiple candidates and determines the intra prediction mode of the current block based on the list of candidates and index information, the information By specifying a candidate index included in the list of candidates, a maximum number of candidates that can be included in the list of candidates is more than three,
the intra-determined prediction mode is one of some extended intra prediction modes, and
Intra extended prediction modes include a flat mode, a DC mode and more than 33 directional prediction modes.
[7]
7. The device of claim 5, wherein the prediction module obtains a prediction sample of the current block based on the intra prediction mode and the reference sample, and corrects the prediction sample obtained using the differential information of the neighboring samples of the current block.
[8]
8. The device of claim 5, wherein the prediction module obtains a prediction sample of the current block based on the intra prediction mode and the reference sample, and corrects the prediction sample obtained based on at least one between a predetermined weight and a default offset.
[9]
9. A procedure for encoding a video, the procedure comprising: determining an intra prediction mode of a current block;
obtain a reference sample for the intra prediction of the current block; Y
perform the intra prediction of the current block based on the intra prediction mode and
the reference sample,
in which the current block is a partitioned coding block based on at least
39
5
10
fifteen
twenty
25
30
35
one of a square tree or a binary tree, and is a non-square block.
[10]
10. The method of claim 9, wherein the determination of the intra prediction mode comprises:
generate a list of candidates that includes multiple candidates; Y
determine the intra prediction mode of the current block based on the candidate list and index information, the index information specifying a candidate index included in the candidate list,
in which a maximum number of candidates that can be included in the list of candidates is more than three,
the intra-determined prediction mode is one of some extended intra prediction modes, and
Intra extended prediction modes include a flat mode, a DC mode and more than 33 directional prediction modes.
[11]
11. The method of claim 9, wherein the realization of the intra prediction comprises:
obtain a prediction sample of the current block based on the intra prediction mode and the reference sample; Y
correct the prediction sample obtained using the differential information of the neighboring samples of the current block.
[12]
12. The method of claim 9, wherein performing the intra prediction comprises:
obtain a prediction sample of the current block based on the intra prediction mode and the reference sample; Y
correct the prediction sample obtained based on at least one of a predetermined weight and a predetermined displacement.
[13]
13. A device for encoding a video, the device comprising:
a prediction module that determines an intra prediction mode of a current block, that obtains a reference sample for the intra prediction of the current block and that makes an intra prediction of the current block based on the intra prediction mode and the sample of reference,
wherein the current block is a partitioned codification block based on at least one of a square tree and a binary tree, and is a non-square block.
40
[14]
14. The device of claim 13, wherein the prediction module generates a list of candidates that includes multiple candidates and determines the intra prediction mode of the current block based on the list of candidates and index information, the
5 index information specifying a candidate index included in the list of candidates,
a maximum number of candidates that can be included in the list of candidates is more than three,
the intra-determined prediction mode is one of some extended intra prediction modes, and
10 intra-extended prediction modes include a flat mode, a DC mode and more than 33 directional prediction modes.
[15]
15. The device of claim 13, wherein the prediction module obtains a prediction sample of the current block based on the intra prediction mode and the
15 reference sample and correct the prediction sample obtained using the differential information of the neighboring samples of the current block.
类似技术:
公开号 | 公开日 | 专利标题
ES2699723B2|2020-10-16|METHOD AND APPARATUS TO TREAT A VIDEO SIGNAL
ES2677193B1|2019-06-19|Procedure and device to process video signals
ES2737841B1|2021-07-27|Method and apparatus for processing video signals
ES2710234B1|2020-03-09|Procedure and device for processing video signals
ES2710807B1|2020-03-27|METHOD AND APPARATUS FOR PROCESSING VIDEO SIGNALS
ES2703607B2|2021-05-13|Method and apparatus for processing video signals
ES2688624A2|2018-11-05|Method and apparatus for processing video signal
ES2724570A2|2019-09-12|Method and apparatus for processing video signals |
ES2784198T3|2020-09-23|Method for block partition and decoding device
ES2737874B2|2020-10-16|METHOD AND APPARATUS FOR PROCESSING VIDEO SIGNAL
ES2711474A2|2019-05-03|Method and device for processing video signal
ES2705251A2|2019-03-22|Method for inducing a fusion candidate block and device using the same |
ES2737843B2|2021-07-15|METHOD AND APPARATUS TO PROCESS A VIDEO SIGNAL
ES2711223A2|2019-04-30|Method and device for processing video signal
ES2737845A2|2020-01-16|METHOD AND APPLIANCE FOR PROCESSING VIDEO SIGNAL |
ES2711473A2|2019-05-03|Method and apparatus for processing video signal
ES2711209A2|2019-04-30|Method and device for processing video signal
同族专利:
公开号 | 公开日
GB2557809B|2021-12-01|
CA2997097A1|2017-03-09|
US20180255299A1|2018-09-06|
US10750174B2|2020-08-18|
ES2719132R1|2019-07-19|
GB2557809A|2018-06-27|
GB201805036D0|2018-05-09|
ES2677193B1|2019-06-19|
GB2596767A|2022-01-05|
EP3343926A1|2018-07-04|
ES2719132A2|2019-07-08|
KR20170026276A|2017-03-08|
EP3343926A4|2019-01-30|
ES2719132B1|2020-05-05|
GB202114959D0|2021-12-01|
US20210067780A1|2021-03-04|
US20210067781A1|2021-03-04|
US20200275100A1|2020-08-27|
CN108353185A|2018-07-31|
ES2677193R1|2018-10-09|
US20210067779A1|2021-03-04|
WO2017039256A1|2017-03-09|
引用文献:
公开号 | 申请日 | 公开日 | 申请人 | 专利标题

KR101365570B1|2007-01-18|2014-02-21|삼성전자주식회사|Method and apparatus for encoding and decoding based on intra prediction|
EP2232877B1|2008-01-10|2018-08-08|Thomson Licensing DTV|Methods and apparatus for illumination compensation of intra-predicted video|
KR101033769B1|2009-05-11|2011-05-09|선문대학교 산학협력단|The method for deciding intra prediction mode of image data|
US9083974B2|2010-05-17|2015-07-14|Lg Electronics Inc.|Intra prediction modes|
KR102043218B1|2010-05-25|2019-11-11|엘지전자 주식회사|New planar prediction mode|
KR101373814B1|2010-07-31|2014-03-18|엠앤케이홀딩스 주식회사|Apparatus of generating prediction block|
WO2012134046A2|2011-04-01|2012-10-04|주식회사 아이벡스피티홀딩스|Method for encoding video|
US9532058B2|2011-06-03|2016-12-27|Qualcomm Incorporated|Intra prediction mode coding with directional partitions|
KR101876173B1|2011-06-17|2018-07-09|엘지전자 주식회사|Method and apparatus for encoding/decoding video in intra prediction mode|
WO2013023518A1|2011-08-17|2013-02-21|Mediatek Singapore Pte. Ltd.|Method and apparatus for intra prediction using non-square blocks|
US9787982B2|2011-09-12|2017-10-10|Qualcomm Incorporated|Non-square transform units and prediction units in video coding|
EP2756678B1|2011-09-15|2017-04-19|VID SCALE, Inc.|Systems and methods for spatial prediction|
KR20130037161A|2011-10-05|2013-04-15|한국전자통신연구원|Method and apparatus of improved inter-layer motion prediction for scalable video coding|
KR101947657B1|2011-10-24|2019-02-14|대가람|Method and apparatus for encoding intra prediction information|
BR112014011123A2|2011-11-08|2017-05-16|Kt Corp|coefficient scan method and apparatus based on prediction unit partition mode|
US20130163664A1|2011-12-22|2013-06-27|Qualcomm Incorporated|Unified partition mode table for intra-mode coding|
EP2800372A4|2011-12-30|2015-12-09|Humax Holdings Co Ltd|Method and device for encoding three-dimensional image, and decoding method and device|
CN109218730A|2012-01-19|2019-01-15|华为技术有限公司|Reference pixel reduction for LM intra prediction|
RS60786B1|2012-01-20|2020-10-30|Dolby Laboratories Licensing Corp|Intra prediction mode mapping method|
KR101620619B1|2012-01-30|2016-05-12|한국전자통신연구원|Method for encoding/decoding of intra prediction mode and apparatus thereof|
KR20140008503A|2012-07-10|2014-01-21|한국전자통신연구원|Method and apparatus for image encoding/decoding|
US9906786B2|2012-09-07|2018-02-27|Qualcomm Incorporated|Weighted prediction mode for scalable video coding|
US9813709B2|2012-09-28|2017-11-07|Nippon Telegraph And Telephone Corporation|Intra-prediction encoding method, intra-prediction decoding method, intra-prediction encoding apparatus, intra-prediction decoding apparatus, program therefor and recording medium having program recorded thereon|
CN103067716B|2013-01-10|2016-06-29|华为技术有限公司|The decoding method of depth image and coding and decoding device|
US9374578B1|2013-05-23|2016-06-21|Google Inc.|Video coding using combined inter and intra predictors|
US9497473B2|2013-10-03|2016-11-15|Qualcomm Incorporated|High precision explicit weighted prediction for video coding|
US9609343B1|2013-12-20|2017-03-28|Google Inc.|Video coding using compound prediction|
US10455249B2|2015-03-20|2019-10-22|Qualcomm Incorporated|Downsampling process for linear model prediction mode|
EP3306920A4|2015-06-05|2019-01-16|Intellectual Discovery Co., Ltd.|Method and device for encoding and decoding intra-frame prediction|
CN108141604A|2015-06-05|2018-06-08|英迪股份有限公司|Image coding and decoding method and image decoding apparatus|
US10038917B2|2015-06-12|2018-07-31|Microsoft Technology Licensing, Llc|Search strategies for intra-picture prediction modes|
US20160373782A1|2015-06-18|2016-12-22|Qualcomm Incorporated|Intra prediction and intra mode coding|
WO2017043949A1|2015-09-11|2017-03-16|주식회사 케이티|Method and device for processing video signal|WO2018212577A1|2017-05-17|2018-11-22|주식회사 케이티|Method and device for video signal processing|
WO2018216862A1|2017-05-24|2018-11-29|엘지전자 주식회사|Method and device for decoding image according to intra prediction in image coding system|
WO2019009620A1|2017-07-04|2019-01-10|엘지전자 주식회사|Image processing method on basis of intra prediction mode and apparatus therefor|
WO2019083284A1|2017-10-24|2019-05-02|주식회사 윌러스표준기술연구소|Video signal processing method and apparatus|
CN111869211A|2018-03-22|2020-10-30|华为技术有限公司|Image encoding device and method|
KR20210158432A|2018-05-12|2021-12-30|주식회사 윌러스표준기술연구소|Video signal processing method and device using reference sample|
KR20210020137A|2018-06-29|2021-02-23|후아웨이 테크놀러지 컴퍼니 리미티드|Apparatus and method for intra prediction of prediction blocks of video images|
WO2020056759A1|2018-09-21|2020-03-26|Alibaba Group Holding Limited|Method, apparatus, and computer-readable storage medium for small size block coding|
US11240497B2|2018-11-15|2022-02-01|Telefonaktiebolaget Lm Ericsson |Deblocking in a video encoder and/or video decoder|
CN113709459A|2019-01-02|2021-11-26|Oppo广东移动通信有限公司|Intra-frame prediction method, device and computer storage medium|
法律状态:
2018-07-30| BA2A| Patent application published|Ref document number: 2677193 Country of ref document: ES Kind code of ref document: A2 Effective date: 20180730 |
2018-10-09| EC2A| Search report published|Ref document number: 2677193 Country of ref document: ES Kind code of ref document: R1 Effective date: 20181002 |
2019-06-19| FG2A| Definitive protection|Ref document number: 2677193 Country of ref document: ES Kind code of ref document: B1 Effective date: 20190619 |
优先权:
申请号 | 申请日 | 专利标题
KR10-2015-0121629|2015-08-28|
KR20150121629|2015-08-28|
KR10-2015-0122976|2015-08-31|
KR20150122975|2015-08-31|
KR10-2015-0122975|2015-08-31|
KR20150122976|2015-08-31|
PCT/KR2016/009574|WO2017039256A1|2015-08-28|2016-08-29|Method and device for processing video signal|
[返回顶部]